Optimization of High Order Perceptrons

نویسندگان

  • Georg Thimm
  • Bernard Werber
چکیده

Neural networks are widely applied in research and industry However their broader ap plication is hampered by various technical details Among these details are several training parameters and the choice of the topology of the network The subject of this dissertation is therefore the elimination and determination of usually user speci ed learning parameters Furthermore suitable application domains for neural networks are discussed Among all training parameters special attention is given to the learning rate the gain of the sigmoidal function and the initial weight range A theorem is proven which permits the elimination of one of these parameters Furthermore it is shown that for high order perceptrons very small random initial weights are usually optimal in terms of training time and generalization Another important problem in the application of neural networks is to nd a network topology that suits a given data set This favors high order perceptrons over several other neural network architectures as they do not require layers of hidden neurons However the order and the connectivity of a network have to be determined which is possible by two approaches The rst is to remove connections from an initially big network while training it The other approach is to increase gradually the network size Both types of approaches are studied corresponding algorithms are developed and applied to high order perceptrons The dis advantages of both approaches are gone into and their performance experimentally compared Then an outlook on future research on the interpretation and analysis of high order perceptrons and their feasibility is given Finally high order perceptrons and the developed algorithms are applied to a number of real world applications and in order to show their e ciency the obtained performances are compared to those of other approaches

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Bounds on the Degree of High Order Binary Perceptrons Bounds on the Degree of High Order Binary Perceptrons

High order perceptrons are often used in order to reduce the size of neural networks. The complexity of the architecture of a usual multilayer network is then turned into the complexity of the functions performed by each high order unit and in particular by the degree of their polynomials. The main result of this paper provides a bound on the degree of the polynomial of a high order perceptron,...

متن کامل

Bounds on the degree of high order binary perceptrons

High order perceptrons are often used in order to reduce the size of neural networks The complexity of the architecture of a usual multilayer network is then turned into the complexity of the functions performed by each high order unit and in particular by the degree of their polynomials The main result of this paper provides a bound on the degree of the polynomial of a high order perceptron wh...

متن کامل

Hyperparameter Optimization with Factorized Multilayer Perceptrons

In machine learning, hyperparameter optimization is a challenging task that is usually approached by experienced practitioners or in a computationally expensive brute-force manner such as grid-search. Therefore, recent research proposes to use observed hyperparameter performance on already solved problems (i.e. data sets) in order to speed up the search for promising hyperparameter configuratio...

متن کامل

Adaptive High Order Neural Trees for Pattern Recognition

In this paper, a new classifier, called adaptive high order neural tree (AHNT), is proposed for pattern recognition applications. It is a hierarchical multi-level neural network, in which the nodes are organized into a tree topology. It successively partitions the training set into subsets, assigning each subset to a different child node. Each node can be a first-order or a high order perceptro...

متن کامل

The Design and Complexity of Exact Multilayered Perceptrons

We investigate the network complexity of multi-layered perceptrons for solving exactly a given problem. We limit our study to the class of combinatorial optimization problems. It is shown how these problems can be reformulated as binary classification problems and how they can be solved by multi-layered perceptrons.

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2003